Goto

Collaborating Authors

 transforming object


Reviews: Image Captioning: Transforming Objects into Words

Neural Information Processing Systems

Summary - The proposed approach to image captioning extends two prior works, object-based Up-Down method of [2] and Transformer of [22] (already used for image captioning in [21]). Specifically, the authors integrate spatial relations between objects in the captioning Transformer model, proposing the Object Relation Transformer. The modification amounts to introducing an object relation module [9] into the encoding layer of the Transformer model. Tests of statistical significance show that the proposed model outperforms the standard Transformer in terms of CIDEr-D, BLEU-1 and ROUGE-L, while SPICE-attribute breakdown shows improvement for Relation and Count categories. Qualitative results include examples where Object Relation Transformer leads to more correct spatial Relation and Count predictions.


Reviews: Image Captioning: Transforming Objects into Words

Neural Information Processing Systems

An object relation module is included into the transformer model. Improvements are demonstrated using this approach. After reading the rebuttal the reviewers agreed that this is an interesting direction to pursue. The reviewers liked the method and partly the results presented in the rebuttal. However the reviewers also remained concerned that additional evidence is necessary (e.g., proper evaluation on test server, experimentation with different spatial features, more in-depth discussion of the attention visualizations, empirical comparison to prior work and human evaluation).


Image Captioning: Transforming Objects into Words

Neural Information Processing Systems

Image captioning models typically follow an encoder-decoder architecture which uses abstract image feature vectors as input to the encoder. One of the most successful algorithms uses feature vectors extracted from the region proposals obtained from an object detector. In this work we introduce the Object Relation Transformer, that builds upon this approach by explicitly incorporating information about the spatial relationship between input detected objects through geometric attention. Quantitative and qualitative results demonstrate the importance of such geometric attention for image captioning, leading to improvements on all common captioning metrics on the MS-COCO dataset.


Image Captioning: Transforming Objects into Words

Herdade, Simao, Kappeler, Armin, Boakye, Kofi, Soares, Joao

Neural Information Processing Systems

Image captioning models typically follow an encoder-decoder architecture which uses abstract image feature vectors as input to the encoder. One of the most successful algorithms uses feature vectors extracted from the region proposals obtained from an object detector. In this work we introduce the Object Relation Transformer, that builds upon this approach by explicitly incorporating information about the spatial relationship between input detected objects through geometric attention. Quantitative and qualitative results demonstrate the importance of such geometric attention for image captioning, leading to improvements on all common captioning metrics on the MS-COCO dataset. Papers published at the Neural Information Processing Systems Conference.